210 research outputs found

    Customizing GermaNet for the use in deep linguistic processing

    Get PDF
    In this paper we show an approach to the customization of GermaNet to the German HPSG grammar lexicon developed in the Verbmobil project. GermaNet has a broad coverage of the German base vocabulary and fine-grained semantic classification; while the HPSG grammar lexicon is comparatively small und has a coarse-grained semantic classification. In our approach, we have developed a mapping algorithm to relate the synsets in GermaNet with the semantic sorts in HPSG. The evaluation result shows that this approach is useful for the lexical extension of our deep grammar development to cope with real-world text understanding

    Linking flat predicate argument structures

    Get PDF
    This report presents an approach to enriching flat and robust predicate argument structures with more fine-grained semantic information, extracted from underspecified semantic representations and encoded in Minimal Recursion Semantics (MRS). Such representations are provided by a hand-built HPSG grammar with a wide linguistic coverage. A specific semantic representation, called linked predicate argument structure (LPAS), has been worked out, which describes the explicit embedding relationships among predicate argument structures. LPAS can be used as a generic interface language for integrating semantic representations with different granularities. Some initial experiments have been conducted to convert MRS expressions into LPASs. A simple constraint solver is developed to resolve the underspecified dominance relations between the predicates and their arguments in MRS expressions. LPASs are useful for high-precision information extraction and question answering tasks because of their fine-grained semantic structures. In addition, I have attempted to extend the lexicon of the HPSG English Resource Grammar (ERG) exploiting WordNet and to disambiguate the readings of HPSG parsing with the help of a probabilistic parser, in order to process texts from application domains. Following the presented approach, the HPSG ERG grammar can be used for annotating some standard treebank, e.g., the Penn Treebank, with its fine-grained semantics. In this vein, I point out opportunities for a fruitful cooperation of the HPSG annotated Redwood Treebank and the Penn PropBank. In my current work, I exploit HPSG as an additional knowledge resource for the automatic learning of LPASs from dependency structures

    Covid-19 Diagnosis Based on CT Images Through Deep Learning and Data Augmentation

    Get PDF
    Coronavirus disease 2019(Covid-19) has made people around the world suffer. And there are many researchers make efforts on deep learning methods based on CT imgaes, but the limitation of  this work is the lackage of the dataset, which is not easy to obtain. In this study, we try to use data augmentation to compensate this weakness. In the first part, we use traditional DenseNet-169, and the result shows that data augmentation can help improve the calculating speed and the accuracy. In the second part, we combine Self-trans and DenseNet-169, and the result shows that when doing data augmentation, many model performance metrics have been improved. In the third part, we use UNet++, which reaches accuracy of 0.8645. Apart from this, we think GAN and CNN may also make difference

    An integrated architecture for shallow and deep processing

    Get PDF
    We present an architecture for the integration of shallow and deep NLP components which is aimed at flexible combination of different language technologies for a range of practical current and future applications. In particular, we describe the integration of a high-level HPSG parsing system with different high-performance shallow components, ranging from named entity recognition to chunk parsing and shallow clause recognition. The NLP components enrich a representation of natural language text with layers of new XML meta-information using a single shared data structure, called the text chart. We describe details of the integration methods, and show how information extraction and language checking applications for realworld German text benefit from a deep grammatical analysis

    Segmental abnormalities of superior longitudinal fasciculus microstructure in patients with schizophrenia, bipolar disorder, and attention-deficit/hyperactivity disorder: An automated fiber quantification tractography study

    Get PDF
    IntroductionSuperior longitudinal fasciculus (SLF) is a white matter (WM) tract that connects the frontal, parietal and temporal lobes. SLF integrity has been widely assessed in neuroimaging studies of psychiatric disorders, such as schizophrenia (SZ), bipolar disorder (BD), and attention-deficit/hyperactivity disorder (ADHD). However, prior studies have revealed inconsistent findings and comparisons across disorders have not been fully examined.MethodsHere, we obtained data for 113 patients (38 patients with SZ, 40 with BD, 35 with ADHD) and 94 healthy controls from the UCLA Consortium for Neuropsychiatric Phenomic LA5c dataset. We assessed the integrity of 20 major WM tracts with a novel segmentation method by automating fiber tract quantification (AFQ). The AFQ divides each tract into 100 equal parts along the direction of travel, with fractional anisotropy (FA) of each part taken as a characteristic. Differences in FA among the four groups were examined.ResultsCompared to healthy controls, patients with SZ showed significantly lower FA in the second half (51–100 parts) of the SLF. No differences were found between BD and healthy controls, nor between ADHD and healthy controls. Results also demonstrated that patients with SZ showed FA reduction in the second half of the SLF relative to patients with BP. Moreover, greater FA in patients in SLF was positively correlated with the manic-hostility score of the Brief Psychiatry Rating scale.DiscussionThese findings indicated that differences in focal changes in SLF might be a key neurobiological abnormality contributing to characterization of these psychiatric disorders

    Corpora and evaluation tools for multilingual named entity grammar development

    Get PDF
    We present an effort for the development of multilingual named entity grammars in a unification-based finite-state formalism (SProUT). Following an extended version of the MUC7 standard, we have developed Named Entity Recognition grammars for German, Chinese, Japanese, French, Spanish, English, and Czech. The grammars recognize person names, organizations, geographical locations, currency, time and date expressions. Subgrammars and gazetteers are shared as much as possible for the grammars of the different languages. Multilingual corpora from the business domain are used for grammar development and evaluation. The annotation format (named entity and other linguistic information) is described. We present an evaluation tool which provides detailed statistics and diagnostics, allows for partial matching of annotations, and supports user-defined mappings between different annotation and grammar output formats
    corecore